Investigation of Pre-Trained Bidirectional Encoder Representations from Transformers Checkpoints for Indonesian Abstractive Text Summarization

نویسندگان

چکیده

Text summarization aims to reduce text by removing less useful information obtain quickly and precisely. In Indonesian abstractive summarization, the research mostly focuses on multi-document which methods will not work optimally in single-document summarization. As public datasets works English are focusing this study emphasized Abstractive studies frequently use Bidirectional Encoder Representations from Transformers (BERT), since BERT checkpoint is available, it was employed study. This investigated of IndoSum dataset using BERTSum model. The investigation proceeded various combinations model encoders, embedding sizes, decoders. Evaluation results showed that models with more size used Generative Pre-Training (GPT)-like decoder could improve Recall-Oriented Understudy for Gisting (ROUGE) score BERTScore results.

برای دانلود باید عضویت طلایی داشته باشید

برای دانلود متن کامل این مقاله و بیش از 32 میلیون مقاله دیگر ابتدا ثبت نام کنید

اگر عضو سایت هستید لطفا وارد حساب کاربری خود شوید

منابع مشابه

Text Generation for Abstractive Summarization

We have begun work on a framework for abstractive summarization and decided to focus on a module for text generation. For TAC 2010, we thus move away from sentence extraction. Each sentence in the summary we generate is based on a document sentence but it usually contains a smaller amount of information and uses fewer words. The system uses the output of a syntactic parser for a sentence and th...

متن کامل

Neural Abstractive Text Summarization

Abstractive text summarization is a complex task whose goal is to generate a concise version of a text without necessarily reusing the sentences from the original source, but still preserving the meaning and the key contents. We address this issue by modeling the problem as a sequence to sequence learning and exploiting Recurrent Neural Networks (RNNs). This work is a discussion about our ongoi...

متن کامل

Decoupling Encoder and Decoder Networks for Abstractive Document Summarization

Abstractive document summarization seeks to automatically generate a summary for a document, based on some abstract “understanding” of the original document. State-of-the-art techniques traditionally use attentive encoder–decoder architectures. However, due to the large number of parameters in these models, they require large training datasets and long training times. In this paper, we propose ...

متن کامل

Generative Adversarial Network for Abstractive Text Summarization

In this paper, we propose an adversarial process for abstractive text summarization, in which we simultaneously train a generative model G and a discriminative model D. In particular, we build the generator G as an agent of reinforcement learning, which takes the raw text as input and predicts the abstractive summarization. We also build a discriminator which attempts to distinguish the generat...

متن کامل

Framework for Abstractive Summarization using Text-to-Text Generation

We propose a new, ambitious framework for abstractive summarization, which aims at selecting the content of a summary not from sentences, but from an abstract representation of the source documents. This abstract representation relies on the concept of Information Items (INIT), which we define as the smallest element of coherent information in a text or a sentence. Our framework differs from pr...

متن کامل

ذخیره در منابع من


  با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید

ژورنال

عنوان ژورنال: Journal of ICT

سال: 2021

ISSN: ['1675-414X', '2180-3862']

DOI: https://doi.org/10.32890/jict2022.21.1.4